near-minimax-optimal distributional reinforcement learning
Near-Minimax-Optimal Distributional Reinforcement Learning with a Generative Model
We propose a new algorithm for model-based distributional reinforcement learning (RL), and prove that it is minimax-optimal for approximating return distributions in the generative model regime (up to logarithmic factors), the first result of this kind for any distributional RL algorithm. Our analysis also provides new theoretical perspectives on categorical approaches to distributional RL, as well as introducing a new distributional Bellman equation, the stochastic categorical CDF Bellman equation, which we expect to be of independent interest. Finally, we provide an experimental study comparing a variety of model-based distributional RL algorithms, with several key takeaways for practitioners.
Near-Minimax-Optimal Distributional Reinforcement Learning with a Generative Model
Rowland, Mark, Wenliang, Li Kevin, Munos, Rémi, Lyle, Clare, Tang, Yunhao, Dabney, Will
We propose a new algorithm for model-based distributional reinforcement learning (RL), and prove that it is minimax-optimal for approximating return distributions with a generative model (up to logarithmic factors), resolving an open question of Zhang et al. (2023). Our analysis provides new theoretical results on categorical approaches to distributional RL, and also introduces a new distributional Bellman equation, the stochastic categorical CDF Bellman equation, which we expect to be of independent interest. We also provide an experimental study comparing several model-based distributional RL algorithms, with several takeaways for practitioners.
- Research Report > New Finding (0.34)
- Research Report > Experimental Study (0.34)